guiding diffusion model
Guiding Diffusion Models for Versatile Face Restoration via Partial Guidance - Supplementary Material - Anonymous Author(s) Affiliation Address email
In this supplementary material, we provide additional discussions and results. B, we provide more results on various tasks, i.e., blind face restoration, old photo restoration, During the inference process, there involves hyperparameters belonging to three categories. Parameters for optional quality enhancement ( e.g., the range for multiple gradient steps to take place Table 1: Default hyperparameter settings in our experiments.T ask Sampling Partial Guidance Optional s As shown in Fig.1, when all the other inference settings are the same, we Input faces are corrupted by real-world degradations. Our method produces high-quality faces with faithful details. ( Zoom in for best view) 3 B.2 More Results on Old Photo Restoration This work focuses on restoring images corrupted by various forms of degradations. This could potentially lead to deceptive information, such as incorrect identity recognition.
PGDiff: Guiding Diffusion Models for Versatile Face Restoration via Partial Guidance
Exploiting pre-trained diffusion models for restoration has recently become a favored alternative to the traditional task-specific training approach. Previous works have achieved noteworthy success by limiting the solution space using explicit degradation models. However, these methods often fall short when faced with complex degradations as they generally cannot be precisely modeled. In this paper, we introduce $\textit{partial guidance}$, a fresh perspective that is more adaptable to real-world degradations compared to existing works. Rather than specifically defining the degradation process, our approach models the desired properties, such as image structure and color statistics of high-quality images, and applies this guidance during the reverse diffusion process. These properties are readily available and make no assumptions about the degradation process. When combined with a diffusion prior, this partial guidance can deliver appealing results across a range of restoration tasks. Additionally, our method can be extended to handle composite tasks by consolidating multiple high-quality image properties, achieved by integrating the guidance from respective tasks. Experimental results demonstrate that our method not only outperforms existing diffusion-prior-based approaches but also competes favorably with task-specific models.
Smoothed Energy Guidance: Guiding Diffusion Models with Reduced Energy Curvature of Attention
Conditional diffusion models have shown remarkable success in visual content generation, producing high-quality samples across various domains, largely due to classifier-free guidance (CFG). Recent attempts to extend guidance to unconditional models have relied on heuristic techniques, resulting in suboptimal generation quality and unintended effects. In this work, we propose Smoothed Energy Guidance (SEG), a novel training- and condition-free approach that leverages the energy-based perspective of the self-attention mechanism to enhance image generation. By defining the energy of self-attention, we introduce a method to reduce the curvature of the energy landscape of attention and use the output as the unconditional prediction. Practically, we control the curvature of the energy landscape by adjusting the Gaussian kernel parameter while keeping the guidance scale parameter fixed.
PGDiff: Guiding Diffusion Models for Versatile Face Restoration via Partial Guidance
Exploiting pre-trained diffusion models for restoration has recently become a favored alternative to the traditional task-specific training approach. Previous works have achieved noteworthy success by limiting the solution space using explicit degradation models. However, these methods often fall short when faced with complex degradations as they generally cannot be precisely modeled. In this paper, we introduce \textit{partial guidance}, a fresh perspective that is more adaptable to real-world degradations compared to existing works. Rather than specifically defining the degradation process, our approach models the desired properties, such as image structure and color statistics of high-quality images, and applies this guidance during the reverse diffusion process. These properties are readily available and make no assumptions about the degradation process.